1,546 research outputs found
Deep Trans-layer Unsupervised Networks for Representation Learning
Learning features from massive unlabelled data is a vast prevalent topic for
high-level tasks in many machine learning applications. The recent great
improvements on benchmark data sets achieved by increasingly complex
unsupervised learning methods and deep learning models with lots of parameters
usually requires many tedious tricks and much expertise to tune. However,
filters learned by these complex architectures are quite similar to standard
hand-crafted features visually. In this paper, unsupervised learning methods,
such as PCA or auto-encoder, are employed as the building block to learn filter
banks at each layer. The lower layer responses are transferred to the last
layer (trans-layer) to form a more complete representation retaining more
information. In addition, some beneficial methods such as local contrast
normalization and whitening are added to the proposed deep trans-layer networks
to further boost performance. The trans-layer representations are followed by
block histograms with binary encoder schema to learn translation and rotation
invariant representations, which are utilized to do high-level tasks such as
recognition and classification. Compared to traditional deep learning methods,
the implemented feature learning method has much less parameters and is
validated in several typical experiments, such as digit recognition on MNIST
and MNIST variations, object recognition on Caltech 101 dataset and face
verification on LFW dataset. The deep trans-layer unsupervised learning
achieves 99.45% accuracy on MNIST dataset, 67.11% accuracy on 15 samples per
class and 75.98% accuracy on 30 samples per class on Caltech 101 dataset,
87.10% on LFW dataset.Comment: 21 pages, 3 figure
Diffusion Maximum Correntropy Criterion Algorithms for Robust Distributed Estimation
Robust diffusion adaptive estimation algorithms based on the maximum
correntropy criterion (MCC), including adaptation to combination MCC and
combination to adaptation MCC, are developed to deal with the distributed
estimation over network in impulsive (long-tailed) noise environments. The cost
functions used in distributed estimation are in general based on the mean
square error (MSE) criterion, which is desirable when the measurement noise is
Gaussian. In non-Gaussian situations, such as the impulsive-noise case, MCC
based methods may achieve much better performance than the MSE methods as they
take into account higher order statistics of error distribution. The proposed
methods can also outperform the robust diffusion least mean p-power(DLMP) and
diffusion minimum error entropy (DMEE) algorithms. The mean and mean square
convergence analysis of the new algorithms are also carried out.Comment: 17 pages,10 figure
Robust Adaptive Sparse Channel Estimation in the Presence of Impulsive Noises
Broadband wireless channels usually have the sparse nature. Based on the
assumption of Gaussian noise model, adaptive filtering algorithms for
reconstruction sparse channels were proposed to take advantage of channel
sparsity. However, impulsive noises are often existed in many advance broadband
communications systems. These conventional algorithms are vulnerable to
deteriorate due to interference of impulsive noise. In this paper, sign least
mean square algorithm (SLMS) based robust sparse adaptive filtering algorithms
are proposed for estimating channels as well as for mitigating impulsive noise.
By using different sparsity-inducing penalty functions, i.e., zero-attracting
(ZA), reweighted ZA (RZA), reweighted L1-norm (RL1) and Lp-norm (LP), the
proposed SLMS algorithms are termed as SLMS-ZA, SLMS-RZA, LSMS-RL1 and SLMS-LP.
Simulation results are given to validate the proposed algorithms.Comment: 5 pages, 4 figures, submitted for DSP2015 conference pape
DRPose3D: Depth Ranking in 3D Human Pose Estimation
In this paper, we propose a two-stage depth ranking based method (DRPose3D)
to tackle the problem of 3D human pose estimation. Instead of accurate 3D
positions, the depth ranking can be identified by human intuitively and learned
using the deep neural network more easily by solving classification problems.
Moreover, depth ranking contains rich 3D information. It prevents the 2D-to-3D
pose regression in two-stage methods from being ill-posed. In our method,
firstly, we design a Pairwise Ranking Convolutional Neural Network (PRCNN) to
extract depth rankings of human joints from images. Secondly, a coarse-to-fine
3D Pose Network(DPNet) is proposed to estimate 3D poses from both depth
rankings and 2D human joint locations. Additionally, to improve the generality
of our model, we introduce a statistical method to augment depth rankings. Our
approach outperforms the state-of-the-art methods in the Human3.6M benchmark
for all three testing protocols, indicating that depth ranking is an essential
geometric feature which can be learned to improve the 3D pose estimation.Comment: Accepted by the 27th International Joint Conference on Artificial
Intelligence (IJCAI 2018
Weakly-Supervised Discovery of Geometry-Aware Representation for 3D Human Pose Estimation
Recent studies have shown remarkable advances in 3D human pose estimation
from monocular images, with the help of large-scale in-door 3D datasets and
sophisticated network architectures. However, the generalizability to different
environments remains an elusive goal. In this work, we propose a geometry-aware
3D representation for the human pose to address this limitation by using
multiple views in a simple auto-encoder model at the training stage and only 2D
keypoint information as supervision. A view synthesis framework is proposed to
learn the shared 3D representation between viewpoints with synthesizing the
human pose from one viewpoint to the other one. Instead of performing a direct
transfer in the raw image-level, we propose a skeleton-based encoder-decoder
mechanism to distil only pose-related representation in the latent space. A
learning-based representation consistency constraint is further introduced to
facilitate the robustness of latent 3D representation. Since the learnt
representation encodes 3D geometry information, mapping it to 3D pose will be
much easier than conventional frameworks that use an image or 2D coordinates as
the input of 3D pose estimator. We demonstrate our approach on the task of 3D
human pose estimation. Comprehensive experiments on three popular benchmarks
show that our model can significantly improve the performance of
state-of-the-art methods with simply injecting the representation as a robust
3D prior.Comment: Accepted as a CVPR 2019 oral paper. Project page:
https://kwanyeelin.github.io
Sparsity Aware Normalized Least Mean p-power Algorithms with Correntropy Induced Metric Penalty
For identifying the non-Gaussian impulsive noise systems, normalized LMP
(NLMP) has been proposed to combat impulsive-inducing instability. However, the
standard algorithm is without considering the inherent sparse structure
distribution of unknown system. To exploit sparsity as well as to mitigate the
impulsive noise, this paper proposes a sparse NLMP algorithm, i.e., Correntropy
Induced Metric (CIM) constraint based NLMP (CIMNLMP). Based on the first
proposed algorithm, moreover, we propose an improved CIM constraint variable
regularized NLMP(CIMVRNLMP) algorithm by utilizing variable regularized
parameter(VRP) selection method which can further adjust convergence speed and
steady-state error. Numerical simulations are given to confirm the proposed
algorithms.Comment: 5 pages, 4 figures, submitted for DSP201
Bias-Compensated Normalized Maximum Correntropy Criterion Algorithm for System Identification with Noisy Input
This paper proposed a bias-compensated normalized maximum correntropy
criterion (BCNMCC) algorithm charactered by its low steady-state misalignment
for system identification with noisy input in an impulsive output noise
environment. The normalized maximum correntropy criterion (NMCC) is derived
from a correntropy based cost function, which is rather robust with respect to
impulsive noises. To deal with the noisy input, we introduce a bias-compensated
vector (BCV) to the NMCC algorithm, and then an unbiasedness criterion and some
reasonable assumptions are used to compute the BCV. Taking advantage of the
BCV, the bias caused by the input noise can be effectively suppressed. System
identification simulation results demonstrate that the proposed BCNMCC
algorithm can outperform other related algorithms with noisy input especially
in an impulsive output noise environment.Comment: 14 pages, 4 figure
An Attention Enhanced Graph Convolutional LSTM Network for Skeleton-Based Action Recognition
Skeleton-based action recognition is an important task that requires the
adequate understanding of movement characteristics of a human action from the
given skeleton sequence. Recent studies have shown that exploring spatial and
temporal features of the skeleton sequence is vital for this task.
Nevertheless, how to effectively extract discriminative spatial and temporal
features is still a challenging problem. In this paper, we propose a novel
Attention Enhanced Graph Convolutional LSTM Network (AGC-LSTM) for human action
recognition from skeleton data. The proposed AGC-LSTM can not only capture
discriminative features in spatial configuration and temporal dynamics but also
explore the co-occurrence relationship between spatial and temporal domains. We
also present a temporal hierarchical architecture to increases temporal
receptive fields of the top AGC-LSTM layer, which boosts the ability to learn
the high-level semantic representation and significantly reduces the
computation cost. Furthermore, to select discriminative spatial information,
the attention mechanism is employed to enhance information of key joints in
each AGC-LSTM layer. Experimental results on two datasets are provided: NTU
RGB+D dataset and Northwestern-UCLA dataset. The comparison results demonstrate
the effectiveness of our approach and show that our approach outperforms the
state-of-the-art methods on both datasets.Comment: Accepted by CVPR201
HFL-RC System at SemEval-2018 Task 11: Hybrid Multi-Aspects Model for Commonsense Reading Comprehension
This paper describes the system which got the state-of-the-art results at
SemEval-2018 Task 11: Machine Comprehension using Commonsense Knowledge. In
this paper, we present a neural network called Hybrid Multi-Aspects (HMA)
model, which mimic the human's intuitions on dealing with the multiple-choice
reading comprehension. In this model, we aim to produce the predictions in
multiple aspects by calculating attention among the text, question and choices,
and combine these results for final predictions. Experimental results show that
our HMA model could give substantial improvements over the baseline system and
got the first place on the final test set leaderboard with the accuracy of
84.13%.Comment: 6 page
Maximum correntropy criterion based sparse adaptive filtering algorithms for robust channel estimation under non-Gaussian environments
Sparse adaptive channel estimation problem is one of the most important
topics in broadband wireless communications systems due to its simplicity and
robustness. So far many sparsity-aware channel estimation algorithms have been
developed based on the well-known minimum mean square error (MMSE) criterion,
such as the zero-attracting least mean square (ZALMS), which are robust under
Gaussian assumption. In non-Gaussian environments, however, these methods are
often no longer robust especially when systems are disturbed by random
impulsive noises. To address this problem, we propose in this work a robust
sparse adaptive filtering algorithm using correntropy induced metric (CIM)
penalized maximum correntropy criterion (MCC) rather than conventional MMSE
criterion for robust channel estimation. Specifically, MCC is utilized to
mitigate the impulsive noise while CIM is adopted to exploit the channel
sparsity efficiently. Both theoretical analysis and computer simulations are
provided to corroborate the proposed methods.Comment: 29 pages, 12 figures, accepted by Journal of the Franklin Institut
- …